Current:Home > StocksPeople are trying to claim real videos are deepfakes. The courts are not amused -InvestPioneer
People are trying to claim real videos are deepfakes. The courts are not amused
Indexbit Exchange View
Date:2025-04-10 16:20:49
In 2016, Elon Musk went on stage at a tech conference outside Los Angeles and made a bold statement about the self-driving capability of Teslas.
"A Model S and Model X at this point can drive autonomously with greater safety than a person. Right now," the CEO told the Code Conference audience during a Q&A session following his interview with tech journalists Walt Mossberg and Kara Swisher.
Video of the session has been up on YouTube for nearly seven years. But it recently came back into the spotlight as part of a lawsuit brought by the family of a man who died when his Tesla crashed while using the self-driving feature. The family's lawyers cited that 2016 claim, along with others Musk has made about Tesla's self-driving software.
But the carmaker's lawyers pushed back.
Musk, "like many public figures, is the subject of many 'deepfake' videos and audio recordings that purport to show him saying and doing things he never actually said or did," they wrote in a court filing, going on to describe several fake videos of the billionaire.
Thanks to advances in artificial intelligence, it's easier than ever to create images and video of things that don't exist, or events that never happened. That's spurring warnings about digital fakery being used to spread propaganda and disinformation, impersonate celebrities and politicians, manipulate elections and scam people.
But the unleashing of powerful generative AI to the public is also raising concerns about another phenomenon: that as the technology becomes more prevalent, it will become easier to claim that anything is fake.
"That's exactly what we were concerned about: that when we entered this age of deepfakes, anybody can deny reality," said Hany Farid, a digital forensics expert and professor at the University of California, Berkeley. "That is the classic liar's dividend."
The liar's dividend is a term coined by law professors Bobby Chesney and Danielle Citron in a 2018 paper laying out the challenges deepfakes present to privacy, democracy, and national security.
The idea is, as people become more aware of how easy it is to fake audio and video, bad actors can weaponize that skepticism.
"Put simply: a skeptical public will be primed to doubt the authenticity of real audio and video evidence," Chesney and Citron wrote.
So far, courts aren't buying claims of deepfaked evidence
In Musk's case, the judge did not buy his lawyers' claims.
"What Tesla is contending is deeply troubling to the Court," Judge Evette Pennypacker wrote in a ruling ordering Musk to testify under oath.
"Their position is that because Mr. Musk is famous and might be more of a target for deep fakes, his public statements are immune," she wrote. "In other words, Mr. Musk, and others in his position, can simply say whatever they like in the public domain, then hide behind the potential for their recorded statements being a deep fake to avoid taking ownership of what they did actually say and do. The Court is unwilling to set such a precedent by condoning Tesla's approach here."
The Tesla lawsuit is not the first time deepfakes have been invoked in an attempt to rebut evidence — and likely won't be the last.
Two of the defendants on trial for the January 6th riot tried to claim video showing them at the Capitol could have been created or manipulated by artificial intelligence. Lawyers for one cited a 2017 deepfake of former President Barack Obama, created by researchers, as reason for the court to doubt footage of the riot that had been taken from YouTube.
But ultimately, both men were found guilty.
"So far, those seem to have been Hail Mary passes that have not succeeded," said Riana Pfefferkorn, a research scholar at the Stanford Internet Observatory who warned about the looming impact of deepfakes in the courtroom in a 2020 law review article.
While the threat that deepfakes could be offered as evidence is real, she said, "the courts have built up hundreds of years of resilience against efforts to introduce fake or tampered with evidence, going all the way back to faking somebody's signature on a handwritten document and going up through typed or mimeographed documents, film, video, Photoshopped evidence, and digital photography and videos."
In fact, Pfefferkorn thinks as AI technology proliferates, it's more likely that courts will confront accusations of fakery against real evidence than attempts to introduce fake evidence.
In those cases, she said, ethics rules and lawyers' professional norms have a role to play. Lawyers are not supposed to put forth frivolous arguments, for example.
"I think attorneys' own sense of self-preservation hopefully will go some distance towards incentivising them to do a little due diligence on the front end," Pfefferkorn said.
Juries may demand more proof
But these standards may need to be updated to specifically address what Loyola Law School professor Rebecca Delfino calls "the deepfake defense."
"Right now, it's sort of like the wild, wild west," she said, where lawyers can say, "'Well, let's run this up the flagpole and see what we can do with it.'"
Even as courts adapt to these challenges, the reverberations from AI fakes will still be felt.
If accusations that evidence is deepfaked become more common, juries may come to expect even more proof that evidence is real, in what Pfefferkorn compares to the "CSI Effect".
"If lawyers start to get juries to demand all the bells and whistles to prove that a piece of evidence is not a fake...that is a way for lawyers and for their clients who are seeking to downplay or dismiss damning evidence against them to essentially run up the bills and make it more expensive, more time-consuming for the other side to get that damning piece of evidence admitted," she said.
That could shut out people who don't have the resources to hire experts.
And whether inside or outside the courtroom, denying that real events actually occurred has corrosive effects. Today, most people carry around devices that can record what's happening around us at a moment's notice, enabling eyewitness accounts of protests, crimes, and public events to be easily shared.
Farid worries about a world in which people no longer believe documented evidence of "police violence, human rights violations, a politician saying something inappropriate or illegal."
"Suddenly there's no more reality," he said. "And that is really worrisome because I don't know how we reason about the world."
veryGood! (11972)
Related
- Federal court filings allege official committed perjury in lawsuit tied to Louisiana grain terminal
- Ishana Night Shyamalan talks debut 'The Watchers,' her iconic dad and his 'cheeky cameos'
- Millie Bobby Brown, Bon Jovi's son and the truth about getting married in your early 20s
- 8 dead, dozens hospitalized after drinking bootleg alcohol in Morocco
- California DMV apologizes for license plate that some say mocks Oct. 7 attack on Israel
- Lakers conduct a public coaching search, considering Redick and Hurley, in hopes of pleasing LeBron
- Video of man pushing Black superintendent at daughter's graduation sparks racism claims
- Travis Kelce Reveals How He's Staying Grounded Amid Taylor Swift Relationship
- The FBI should have done more to collect intelligence before the Capitol riot, watchdog finds
- DNC to unveil new billboard calling Trump a convicted felon
Ranking
- Appeals court scraps Nasdaq boardroom diversity rules in latest DEI setback
- National Donut Day 2024 deals: Get free food at Dunkin', Krispy Kreme, Duck Donuts, Sheetz
- Stanley Cup Final difference-makers: Connor McDavid, Aleksander Barkov among 10 stars to watch
- A look back at D-Day: Why the World War II invasion remains important on its 80th anniversary
- Mets have visions of grandeur, and a dynasty, with Juan Soto as major catalyst
- Watch as huge, 12-foot alligator dangles from grip of grapple truck in Texas
- California made it easier to vote, but some with disabilities still face barriers
- Kevin Costner said he refused to shorten his 17-minute eulogy for Whitney Houston: I was her imaginary bodyguard.
Recommendation
Off the Grid: Sally breaks down USA TODAY's daily crossword puzzle, Triathlon
Powerball winning numbers for June 5 drawing: Jackpot climbs to $206 million
U.S. man who killed girlfriend, stuffed body in suitcase gets 42 years for femicide in Colombia
Today is last day Walmart shoppers can claim up to $500. Here's how.
Opinion: Gianni Infantino, FIFA sell souls and 2034 World Cup for Saudi Arabia's billions
Kevin Costner said he refused to shorten his 17-minute eulogy for Whitney Houston: I was her imaginary bodyguard.
Over 1.2 million rechargeable lights are under recall for fire hazards, following one reported death
This underused Social Security move will boost the average check by $460 in 3 years